Multimodal Adaptive Interfaces

نویسنده

  • Deb Roy
چکیده

Our group is interested in creating human machine interfaces which use natural modalities such as vision and speech to sense and interpret a user's actions [6]. In this paper we describe recent work on multimodal adaptive interfaces which combine automatic speech recognition, computer vision for gesture tracking, and machine learning techniques. Speech is the primary mode of communication between people and should also be used in computer human communication. Gesture usually accompanies speech and provides information which is at times complementary and at times redundant to the information in the speech stream. Depending on the task at hand and the user's preferences, she will use a combination of speech and gesture in di erent ways to communicate her intent. In this paper we present preliminary results of an interface which lets the user communicate using a combination of speech and diectic (pointing) gestures. Although other e orts have been made to build multimodal interfaces, we present a system which centers around on-line learning to actively acquire communicative primitives from interactions with the user. In this paper we begin in Section 2 by examining some of the problems of designing interfaces which use natural modalities and motivate our approach which centers around enabling the interface to learn from the user. In Section 3 we give an overview of our approach to addressing the issues raised in Section 2. Section 4 describes the multimodal sensory environment we have built for developing our interfaces. This environment includes a vision based hand tracking system and a phonetic speech recognizer. In Section 5 we introduce Toco the Toucan, an animated synthetic character which provides embodiment for the interface. Section 6 details the learning algorithm which allows Toco to learn the acoustic models and meanings of words as the user points to virtual objects and talk about them. In Section 7 we summarize our work this far and nally in Section 8 we provide a road map for where this work is going.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Adaptive Mobile Multimodal User Interfaces for Dynamic Accessibility

In this paper we discuss how to exploit solutions able to support mobile multimodal adaptive user interfaces for dynamic accessibility. We focus on an approach based on the use of declarative user interface languages and oriented to Web applications accessed through emerging ubiquitous environments.

متن کامل

Human-centric framework for perceptually adaptive interfaces

Multimodal interfaces have long held the promise of enhanced and effective human machine interaction. The ultimate goal of multimodal interfaces is to facilitate human activity allowing seamless exchange of information. This goal requires a coordinated development effort that incorporates a thorough understanding of human perceptual system in the design of interfaces. In this manner, multimodal...

متن کامل

Adaptive multimodal web user interfaces for smart work environments

This paper presents a solution for supporting adaptive user interfaces in work environments that require operators to move about in dynamic contexts for manipulating various physical objects. The solution architecture is built upon the use of logical languages for interactive applications integrated with context aware and adaptive features. The proposed architecture is able to adapt to specific...

متن کامل

Tangible Objects for the Acquisition of Multimodal Interaction Patterns

Multimodal user interfaces offer more intuitive interaction for end-users, however, usually only through predefined input schemes. This paper describes a user experiment for multimodal interaction pattern identification, using head gesture and speech inputs for a 3D graph manipulation. We show that a direct mapping between head gestures and the 3D object predominates, however even for such a si...

متن کامل

Semantic Modelling in Support of Adaptive Multimodal Interface Design

The design of multimodal interfaces requires intelligent data interpretation in order to guarantee seamless adaptation to the user’s needs and context. HMI (humanmachine interaction) design accommodates varying forms of interaction patterns, depending on what is most appropriate for a particular user at a particular time. These design patterns are a powerful means of documenting reusable design...

متن کامل

A Cross-Layer Design for Adaptive Multimodal Interfaces in Pervasive Computing

Multimodal interfaces have attracted more and more attention. Most researches focus on each communication mode independently and then fuse the information at the application level. Recently, several frameworks and models have been proposed to support the design and development of multimodal interfaces. However, it is still a challenging issue of supporting adaptations in the multimodal interfac...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1997